# Wikipedia Training

Llama 160m
Apache-2.0
This is a LLaMA-like model with only 160 million parameters, trained on data from Wikipedia and parts of the C4-en and C4-realnewslike datasets.
Large Language Model Transformers English
L
JackFram
275.14k
34
Albert Base Japanese V1
MIT
This is a pre-trained Japanese ALBERT model primarily designed for fill-mask tasks, supporting Japanese text processing.
Large Language Model Transformers Japanese
A
ken11
609
0
Nordic Roberta Wiki
A Nordic RoBERTa model trained on Swedish, Danish, and Norwegian Wikipedia, suitable for fill-mask tasks.
Large Language Model Other
N
flax-community
13
0
Tamillion
A Tamil pre-trained model based on the ELECTRA framework, with the second version trained on TPUs and expanded corpus scale
Large Language Model Transformers Other
T
monsoon-nlp
58
2
Bert Base Multilingual Cased Finetuned Dutch Squad2
Dutch Q&A model fine-tuned from Google's multilingual BERT, trained on translated SQuAD2.0 dataset
Question Answering System Other
B
henryk
55
6
Indo Roberta Small
MIT
Indonesian Small RoBERTa is a masked language model based on the RoBERTa architecture, specifically trained for Indonesian language, suitable for text infilling and feature extraction tasks.
Large Language Model Other
I
w11wo
50
1
Distilbert Base Fr Cased
Apache-2.0
This is a French-specific version of the multilingual distilled BERT base model, retaining the original model's accuracy while being more compact.
Large Language Model Transformers French
D
Geotrend
1,104
2
Bert Base En Es Cased
Apache-2.0
This is a streamlined version based on bert-base-multilingual-cased, specializing in English and Spanish processing, fully reproducing the original model's representation output while maintaining its accuracy.
Large Language Model Other
B
Geotrend
13
0
Distilbert Base Th Cased
Apache-2.0
A Thai version customized from the multilingual distilled BERT base model, retaining the original model's accuracy and feature representation capabilities.
Large Language Model Transformers Other
D
Geotrend
50
0
Distilbert Base En De Cased
Apache-2.0
This is a lightweight version of distilbert-base-multilingual-cased, focused on bilingual processing of English and German, maintaining the representation capability and accuracy of the original model.
Large Language Model Transformers Other
D
Geotrend
23
0
Bert Base En Th Cased
Apache-2.0
A customized lightweight language version model based on bert-base-multilingual-cased, supporting English and Thai while maintaining the original model's representation output and accuracy.
Large Language Model Other
B
Geotrend
14
1
Distilbert Base Ru Cased
Apache-2.0
This is a compact version of the multilingual distilled BERT base model (case-sensitive), specifically optimized for Russian, capable of generating semantic representations identical to the original model while maintaining its accuracy.
Large Language Model Transformers Other
D
Geotrend
498
2
Distilbert Base En Ar Cased
Apache-2.0
This is a distilled version of distilbert-base-multilingual-cased, specifically optimized for English and Arabic processing while maintaining the original model's accuracy.
Large Language Model Transformers Other
D
Geotrend
31
0
Distilbert Base Pt Cased
Apache-2.0
This is a compact version of distilbert-base-multilingual-cased, specifically designed for Portuguese, while maintaining the accuracy of the original model.
Large Language Model Transformers Other
D
Geotrend
46
2
Distilbert Base Vi Cased
Apache-2.0
This is a customized distilled version of the multilingual DistilBERT base model, optimized for Vietnamese language processing while retaining the original model's representational capabilities and accuracy.
Large Language Model Transformers Other
D
Geotrend
41
1
Bert Base En It Cased
Apache-2.0
This is a streamlined version of bert-base-multilingual-cased, specifically optimized for English and Italian while retaining the original model's representational capabilities.
Large Language Model Other
B
Geotrend
17
0
Bert Base Zh Cased
Apache-2.0
A streamlined Chinese BERT model customized from bert-base-multilingual-cased, supporting Chinese processing while maintaining original accuracy
Large Language Model Chinese
B
Geotrend
15
0
Bert Base En Uk Cased
Apache-2.0
This is a customized compact version of bert-base-multilingual-cased, specifically supporting English and Ukrainian processing while maintaining the original model's representation output and accuracy.
Large Language Model Other
B
Geotrend
22
0
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase